Certified Generative AI Engineer Associate v1.0

Page:    1 / 3   
Exam contains 45 questions

A Generative Al Engineer interfaces with an LLM with prompt/response behavior that has been trained on customer calls inquiring about product availability. The LLM is designed to output “In Stock” if the product is available or only the term “Out of Stock” if not.
Which prompt will work to allow the engineer to respond to call classification labels correctly?

  • A. Respond with “In Stock” if the customer asks for a product.
  • B. You will be given a customer call transcript where the customer asks about product availability. The outputs are either “In Stock” or “Out of Stock”. Format the output in JSON, for example: {“call_id”: “123”, “label”: “In Stock”}.
  • C. Respond with “Out of Stock” if the customer asks for a product.
  • D. You will be given a customer call transcript where the customer inquires about product availability. Respond with “In Stock” if the product is available or “Out of Stock” if not.


Answer : B

A Generative Al Engineer is tasked with developing a RAG application that will help a small internal group of experts at their company answer specific questions, augmented by an internal knowledge base. They want the best possible quality in the answers, and neither latency nor throughput is a huge concern given that the user group is small and they’re willing to wait for the best answer. The topics are sensitive in nature and the data is highly confidential and so, due to regulatory requirements, none of the information is allowed to be transmitted to third parties.
Which model meets all the Generative Al Engineer’s needs in this situation?

  • A. Dolly 1.5B
  • B. OpenAI GPT-4
  • C. BGE-large
  • D. Llama2-70B


Answer : D

A Generative Al Engineer would like an LLM to generate formatted JSON from emails. This will require parsing and extracting the following information: order ID, date, and sender email. Here’s a sample email:

They will need to write a prompt that will extract the relevant information in JSON format with the highest level of output accuracy.
Which prompt will do that?

  • A. You will receive customer emails and need to extract date, sender email, and order ID. You should return the date, sender email, and order ID information in JSON format.
  • B. You will receive customer emails and need to extract date, sender email, and order ID. Return the extracted information in JSON format.
    Here’s an example: {“date”: “April 16, 2024”, “sender_email”: “[email protected]”, “order_id”: “RE987D”}
  • C. You will receive customer emails and need to extract date, sender email, and order ID. Return the extracted information in a human-readable format.
  • D. You will receive customer emails and need to extract date, sender email, and order IReturn the extracted information in JSON format.


Answer : B

A Generative AI Engineer has been asked to build an LLM-based question-answering application. The application should take into account new documents that are frequently published. The engineer wants to build this application with the least cost and least development effort and have it operate at the lowest cost possible.
Which combination of chaining components and configuration meets these requirements?

  • A. For the application a prompt, a retriever, and an LLM are required. The retriever output is inserted into the prompt which is given to the LLM to generate answers.
  • B. The LLM needs to be frequently with the new documents in order to provide most up-to-date answers.
  • C. For the question-answering application, prompt engineering and an LLM are required to generate answers.
  • D. For the application a prompt, an agent and a fine-tuned LLM are required. The agent is used by the LLM to retrieve relevant content that is inserted into the prompt which is given to the LLM to generate answers.


Answer : A

A Generative AI Engineer is creating an agent-based LLM system for their favorite monster truck team. The system can answer text based questions about the monster truck team, lookup event dates via an API call, or query tables on the team’s latest standings.
How could the Generative AI Engineer best design these capabilities into their system?

  • A. Ingest PDF documents about the monster truck team into a vector store and query it in a RAG architecture.
  • B. Write a system prompt for the agent listing available tools and bundle it into an agent system that runs a number of calls to solve a query.
  • C. Instruct the LLM to respond with “RAG”, “API”, or “TABLE” depending on the query, then use text parsing and conditional statements to resolve the query.
  • D. Build a system prompt with all possible event dates and table information in the system prompt. Use a RAG architecture to lookup generic text questions and otherwise leverage the information in the system prompt.


Answer : B

A Generative AI Engineer has been asked to design an LLM-based application that accomplishes the following business objective: answer employee HR questions using HR PDF documentation.
Which set of high level tasks should the Generative AI Engineer's system perform?

  • A. Calculate averaged embeddings for each HR document, compare embeddings to user query to find the best document. Pass the best document with the user query into an LLM with a large context window to generate a response to the employee.
  • B. Use an LLM to summarize HR documentation. Provide summaries of documentation and user query into an LLM with a large context window to generate a response to the user.
  • C. Create an interaction matrix of historical employee questions and HR documentation. Use ALS to factorize the matrix and create embeddings. Calculate the embeddings of new queries and use them to find the best HR documentation. Use an LLM to generate a response to the employee question based upon the documentation retrieved.
  • D. Split HR documentation into chunks and embed into a vector store. Use the employee question to retrieve best matched chunks of documentation, and use the LLM to generate a response to the employee based upon the documentation retrieved.


Answer : D

Generative AI Engineer at an electronics company just deployed a RAG application for customers to ask questions about products that the company carries. However, they received feedback that the RAG response often returns information about an irrelevant product.
What can the engineer do to improve the relevance of the RAG’s response?

  • A. Assess the quality of the retrieved context
  • B. Implement caching for frequently asked questions
  • C. Use a different LLM to improve the generated response
  • D. Use a different semantic similarity search algorithm


Answer : A

A Generative AI Engineer is developing a chatbot designed to assist users with insurance-related queries. The chatbot is built on a large language model (LLM) and is conversational. However, to maintain the chatbot’s focus and to comply with company policy, it must not provide responses to questions about politics. Instead, when presented with political inquiries, the chatbot should respond with a standard message:
“Sorry, I cannot answer that. I am a chatbot that can only answer questions around insurance.”
Which framework type should be implemented to solve this?

  • A. Safety Guardrail
  • B. Security Guardrail
  • C. Contextual Guardrail
  • D. Compliance Guardrail


Answer : D

A Generative AI Engineer I using the code below to test setting up a vector store:

Assuming they intend to use Databricks managed embeddings with the default embedding model, what should be the next logical function call?

  • A. vsc.get_index()
  • B. vsc.create_delta_sync_index()
  • C. vsc.create_direct_access_index()
  • D. vsc.similarity_search()


Answer : B

A Generative AI Engineer is tasked with deploying an application that takes advantage of a custom MLflow Pyfunc model to return some interim results.
How should they configure the endpoint to pass the secrets and credentials?

  • A. Use spark.conf.set ()
  • B. Pass variables using the Databricks Feature Store API
  • C. Add credentials using environment variables
  • D. Pass the secrets in plain text


Answer : C

A Generative AI Engineer wants to build an LLM-based solution to help a restaurant improve its online customer experience with bookings by automatically handling common customer inquiries. The goal of the solution is to minimize escalations to human intervention and phone calls while maintaining a personalized interaction. To design the solution, the Generative AI Engineer needs to define the input data to the LLM and the task it should perform.
Which input/output pair will support their goal?

  • A. Input: Online chat logs; Output: Group the chat logs by users, followed by summarizing each user’s interactions
  • B. Input: Online chat logs; Output: Buttons that represent choices for booking details
  • C. Input: Customer reviews; Output: Classify review sentiment
  • D. Input: Online chat logs; Output: Cancellation options


Answer : B

What is an effective method to preprocess prompts using custom code before sending them to an LLM?

  • A. Directly modify the LLM’s internal architecture to include preprocessing steps
  • B. It is better not to introduce custom code to preprocess prompts as the LLM has not been trained with examples of the preprocessed prompts
  • C. Rather than preprocessing prompts, it’s more effective to postprocess the LLM outputs to align the outputs to desired outcomes
  • D. Write a MLflow PyFunc model that has a separate function to process the prompts


Answer : D

A Generative AI Engineer is developing an LLM application that users can use to generate personalized birthday poems based on their names.
Which technique would be most effective in safeguarding the application, given the potential for malicious user inputs?

  • A. Implement a safety filter that detects any harmful inputs and ask the LLM to respond that it is unable to assist
  • B. Reduce the time that the users can interact with the LLM
  • C. Ask the LLM to remind the user that the input is malicious but continue the conversation with the user
  • D. Increase the amount of compute that powers the LLM to process input faster


Answer : A

Which indicator should be considered to evaluate the safety of the LLM outputs when qualitatively assessing LLM responses for a translation use case?

  • A. The ability to generate responses in code
  • B. The similarity to the previous language
  • C. The latency of the response and the length of text generated
  • D. The accuracy and relevance of the responses


Answer : D

A Generative AI Engineer is developing a patient-facing healthcare-focused chatbot. If the patient’s question is not a medical emergency, the chatbot should solicit more information from the patient to pass to the doctor’s office and suggest a few relevant pre-approved medical articles for reading. If the patient’s question is urgent, direct the patient to calling their local emergency services.
Given the following user input:
“I have been experiencing severe headaches and dizziness for the past two days.”
Which response is most appropriate for the chatbot to generate?

  • A. Here are a few relevant articles for your browsing. Let me know if you have questions after reading them.
  • B. Please call your local emergency services.
  • C. Headaches can be tough. Hope you feel better soon!
  • D. Please provide your age, recent activities, and any other symptoms you have noticed along with your headaches and dizziness.


Answer : B

Page:    1 / 3   
Exam contains 45 questions

Talk to us!


Have any questions or issues ? Please dont hesitate to contact us

Certlibrary.com is owned by MBS Tech Limited: Room 1905 Nam Wo Hong Building, 148 Wing Lok Street, Sheung Wan, Hong Kong. Company registration number: 2310926
Certlibrary doesn't offer Real Microsoft Exam Questions. Certlibrary Materials do not contain actual questions and answers from Cisco's Certification Exams.
CFA Institute does not endorse, promote or warrant the accuracy or quality of Certlibrary. CFA® and Chartered Financial Analyst® are registered trademarks owned by CFA Institute.
Terms & Conditions | Privacy Policy